Excess risk bounds in robust empirical risk minimization

نویسندگان

چکیده

Abstract This paper investigates robust versions of the general empirical risk minimization algorithm, one core techniques underlying modern statistical methods. Success is based on fact that for a ‘well-behaved’ stochastic process $\left \{ f(X), \ f\in \mathscr F\right \}$ indexed by class functions $f\in F$, averages $\frac{1}{N}\sum _{j=1}^N f(X_j)$ evaluated over sample $X_1,\ldots ,X_N$ i.i.d. copies $X$ provide good approximation to expectations $\mathbb E f(X)$, uniformly large classes F$. However, this might no longer be true if marginal distributions are heavy tailed or contains outliers. We propose version idea replacing proxies and obtain high-confidence bounds excess resulting estimators. In particular, we show estimators can converge $0$ at fast rates with respect size $N$, referring faster than $N^{-1/2}$. discuss implications main results linear logistic regression problems evaluate numerical performance proposed methods simulated real data.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Chaining Bounds for Empirical Risk Minimization

This paper extends the standard chaining technique to prove excess risk upper bounds for empirical risk minimization with random design settings even if the magnitude of the noise and the estimates is unbounded. The bound applies to many loss functions besides the squared loss, and scales only with the sub-Gaussian or subexponential parameters without further statistical assumptions such as the...

متن کامل

Empirical Risk Minimization: Abstract Risk Bounds and Rademacher Averages

1. An abstract framework for ERM To study ERM in a general framework, we will adopt a simplified notation often used in the literature. We have a space Z and a class F of functions f : Z→ [0, 1]. Let P(Z) denote the space of all probability distributions on Z. For each sample size n, the training data are in the form of an n-tuple Zn = (Z1, . . . , Zn) of Z-valued random variables drawn accordi...

متن کامل

Asymptotics in Empirical Risk Minimization

In this paper, we study a two-category classification problem. We indicate the categories by labels Y = 1 and Y = −1. We observe a covariate, or feature, X ∈ X ⊂ R. Consider a collection {ha} of classifiers indexed by a finite-dimensional parameter a, and the classifier ha∗ that minimizes the prediction error over this class. The parameter a∗ is estimated by the empirical risk minimizer ân over...

متن کامل

mixup: BEYOND EMPIRICAL RISK MINIMIZATION

Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple lin...

متن کامل

Aggregation via Empirical Risk Minimization

Given a finite set F of estimators, the problem of aggregation is to construct a new estimator whose risk is as close as possible to the risk of the best estimator in F . It was conjectured that empirical minimization performed in the convex hull of F is an optimal aggregation method, but we show that this conjecture is false. Despite that, we prove that empirical minimization in the convex hul...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Information and Inference: A Journal of the IMA

سال: 2021

ISSN: ['2049-8772', '2049-8764']

DOI: https://doi.org/10.1093/imaiai/iaab004